This article discusses the concept of 'tool masking' as a way to optimize the interaction between LLMs and APIs, arguing that simply exposing all API functionality (as done by MCP) is inefficient and degrades performance. It proposes shaping the tool surface to match the specific use case, improving accuracy, cost, and latency.
This article provides a practical guide to JSON prompting for Large Language Models (LLMs), demonstrating how structuring prompts with JSON improves consistency, accuracy, and scalability. It includes Python coding examples comparing free-form and JSON prompts, and provides access to full code notebooks.
Security researchers have discovered that LLM chatbots can be made to ignore their guardrails by using prompts with terrible grammar and long, run-on sentences. This bypasses safety filters and allows the model to generate harmful responses. The research introduces the 'refusal-affirmation logit gap' as a metric for assessing model vulnerability.
GitHub Models is a suite of developer tools for AI development, offering features like prompt management, model comparison, and quantitative evaluations, integrated directly into GitHub workflows.
A post with pithy observations and clear conclusions from building complex LLM workflows, covering topics like prompt chaining, data structuring, model limitations, and fine-tuning strategies.
This article details new prompting techniques for ChatGPT-4.1, emphasizing structured prompts, precise delimiting, agent creation, long context handling, and chain-of-thought prompting to achieve better results.
This article explains prompt engineering techniques for large language models (LLMs), covering methods like zero-shot, few-shot, system, contextual, role, step-back, chain-of-thought, self-consistency, ReAct, Automatic Prompt Engineering and code prompting. It also details best practices and output configuration for optimal results.
This article details an iterative process of using ChatGPT to explore the parallels between Marvin Minsky's "Society of Mind" and Anthropic's research on Large Language Models, specifically Claude Haiku. The user experimented with different prompts to refine the AI's output, navigating issues like model confusion (GPT-2 vs. Claude) and overly conversational tone. Ultimately, prompting the AI with direct source materials (Minsky’s books and Anthropic's paper) yielded the most insightful analysis, highlighting potential connections like the concept of "A and B brains" within both frameworks.
A guide on implementing prompt engineering patterns to make RAG implementations more effective and efficient, covering patterns like Direct Retrieval, Chain of Thought, Context Enrichment, Instruction-Tuning, and more.
The article discusses how structured, modular software engineering practices enhance the effectiveness of large language models (LLMs) in software development tasks. It emphasizes the importance of clear and coherent code, which allows LLMs to better understand, extend functionality, and debug. The author shares experiences from the Bad Science Fiction project, illustrating how well-engineered code improves AI collaboration.
Key takeaways:
1. **Modular Code**: Use small, well-documented code blocks to aid LLM performance.
2. **Effective Prompts**: Design clear, structured prompts by defining context and refining iteratively.
3. **Chain-of-Thought Models**: Provide precise inputs to leverage structured problem-solving abilities.
4. **Prompt Literacy**: Master expressing computational intent clearly in natural language.
5. **Iterative Refinement**: Utilize AI consultants for continuous code improvement.
6. **Separation of Concerns**: Organize code into server and client roles for better AI interaction.